home
***
CD-ROM
|
disk
|
FTP
|
other
***
search
/
Internet Tools (InfoMagic)
/
Internet Tools.iso
/
ip
/
trace
/
collect
/
RELEASE.Z
/
RELEASE
Wrap
Text File
|
1991-12-03
|
5KB
|
100 lines
Hello,
We need some volunteers to collect wide-area traffic traces. Below we
briefly describe the research context this collection fits in and what you need
to do if you would help us do the collection.
1. Research Context
We are trying to characterize the traffic that flows through the Internet;
you may have seen our SIGCOMM 91 paper on characterizing wide-area TCP/IP
traffic. We are now extending this work to be a complete traffic source model
of the Internet to be used in driving performance evaluations of routing, flow,
congestion, and resource management algorithms.
This type of research requires instrumenting lots of stub networks so that
they save and make available descriptions of all the network traffic that
terminates or originates at their stub. The traffic descriptions we need has
to be much more detailed than the monthly traffic reports NSFNET already makes
available. More specifically, we need sufficient information to describe every
"conversation" with an endpoint at the stub. This requires saving three or
four packet headers per TCP conversation.
This data collection would not interfere with normal operation of the
network and does not pose a network security threat. Sites sensitive to
privacy issues could perform a one-way transformation on the conversation's
source and destination IP addresses.
2. What You Need to Do
If you would like to help us do the collection, you need to do an anonymous
ftp to jerico.usc.edu (128.125.51.6) and get the collection package. The
package is in the directory: ~ftp/pub/jamin/collect. The directory contains
the following two files:
13585 collect.tar.Z
262019 tcpdump-2.0.tar.Z
The file "tcpdump-2.0.tar.Z" is the tcpdump distribution from LBL. You
need this file only if you don't have tcpdump (version 2.0) installed on your
system.
The file "collect.tar.Z" contains the two programs used in the data
collection process: "collect" and "generator."
"Collect" is a shell script which invokes the tcpdump program to collect
network packets. (Tcpdump is a network instrumentation program from Lawrence
Berkeley Lab.) It is important that the collection routine be run on a machine
on the ethernet segment connected to the sites internetwork gateway, so that
all internet packets can be observed.
When running on Ultrix or BSD+BPF, tcpdump can log dropped packets. Not so
on SunOS with the nit interface; the program "generator" is used to estimate
the loss-rate of the collection run. The validity of studies based on
collecting traces depends on the loss-rate of the collection. Since loss-rate
depends on the collection machine's CPU utilization, it is best that the
collection be run on a dedicated workstation. The nit interface of SunOS cannot
collect packets generated by the local machine, thus on SunOS "generator" should
run on a different machine from the one running collect. This means that
estimating the loss-rate of a collection requires three machines: one to run
"collect," one to run "generator," and one to act as the target of packets
generated by "generator" (the load put on the "generator" machines is
negligible). If you absolutely can spare only one machine, you can forgo doing
loss-rate estimation; we would still encourage you to collect data which we will
use to support conclusions drawn from primary traces.
There are more detailed instructions on installing and running the
collection package in the README file and manual pages included in
collect.tar.Z, but to summarize, these are the equipments you need for doing
the collection:
1. A machine running an operating system supported by the tcpdump program, i.e.
one of SunOS, Ultrix, BSD+BPF (See the README file in tcpdump-2.0.tar.Z).
2. The said machine should have enough disk space to hold the collected data.
This space requirement depends on the load on your network. For the three
sites we have done our collection so far, UC Berkeley saw 64,800
conversations during the 24 hour collection period, USC saw 17,950
conversations, and Bellcore saw 14,130 conversations. From these figures,
we calculated that 20MB disk space should be more than enough to hold a
day's worth of data.
And if you are doing loss-rate estimation:
3. A machine to run the "generator" program. Any UNIX machine will do.
4. A machine to act as a target of the "generator" program.
======================================
||
||
+----+
| GW |
+----+
|
|
-------------------------------------------------
| | |
| | |
+---------+ +-------+ +---------+
|generator| |collect| |generator|
| source | | host | | target |
|OPTIONAL | |RQUIRED| |OPTIONAL |
+---------+ +-------+ +---------+
Fig. 1: Example collection run setup.
If you have any problems or questions, please feel free to contact us at:
traffic@excalibur.usc.edu.